knowledge capability
Methodology of Adapting Large English Language Models for Specific Cultural Contexts
Zhang, Wenjing, Xiao, Siqi, Lei, Xuejiao, Wang, Ning, Zhang, Huazheng, An, Meijuan, Yang, Bikun, Liu, Zhaoxiang, Wang, Kai, Lian, Shiguo
The rapid growth of large language models(LLMs) has emerged as a prominent trend in the field of artificial intelligence. However, current state-of-the-art LLMs are predominantly based on English. They encounter limitations when directly applied to tasks in specific cultural domains, due to deficiencies in domain-specific knowledge and misunderstandings caused by differences in cultural values. To address this challenge, our paper proposes a rapid adaptation method for large models in specific cultural contexts, which leverages instruction-tuning based on specific cultural knowledge and safety values data. Taking Chinese as the specific cultural context and utilizing the LLaMA3-8B as the experimental English LLM, the evaluation results demonstrate that the adapted LLM significantly enhances its capabilities in domain-specific knowledge and adaptability to safety values, while maintaining its original expertise advantages.
FoundaBench: Evaluating Chinese Fundamental Knowledge Capabilities of Large Language Models
Li, Wei, Ma, Ren, Wu, Jiang, Gu, Chenya, Peng, Jiahui, Len, Jinyang, Zhang, Songyang, Yan, Hang, Lin, Dahua, He, Conghui
In the burgeoning field of large language models (LLMs), the assessment of fundamental knowledge remains a critical challenge, particularly for models tailored to Chinese language and culture. This paper introduces FoundaBench, a pioneering benchmark designed to rigorously evaluate the fundamental knowledge capabilities of Chinese LLMs. FoundaBench encompasses a diverse array of 3354 multiple-choice questions across common sense and K-12 educational subjects, meticulously curated to reflect the breadth and depth of everyday and academic knowledge. We present an extensive evaluation of 12 state-of-the-art LLMs using FoundaBench, employing both traditional assessment methods and our CircularEval protocol to mitigate potential biases in model responses. Our results highlight the superior performance of models pre-trained on Chinese corpora, and reveal a significant disparity between models' reasoning and memory recall capabilities. The insights gleaned from FoundaBench evaluations set a new standard for understanding the fundamental knowledge of LLMs, providing a robust framework for future advancements in the field.
Bringing Intelligence to Enterprise Content Management, Google Releases Document Understanding AI
At the recent Google Cloud Next Conference, Google announced a new beta machine learning service, called Document Understanding AI. The service targets Enterprise Content Management (ECM) workloads by allowing customers to organize, classify and extract key value pairs from unstructured content, in the enterprise, using Artificial Intelligence (AI) and Machine Learning (ML). Gartner and Forbes estimate that 80% of enterprise data is unstructured and 70% of enterprise data is free-form text like emails, written documents and comments. With the amount of data in enterprise organizations, Google perceives an opportunity to use AI and ML to address enterprise content challenges. Many enterprises see the value in applying AI and machine learning to their business challenges, but not all have the necessary resources to do it.